Relative stability toward diffeomorphisms indicates performance in deep nets*

نویسندگان

چکیده

Abstract Understanding why deep nets can classify data in large dimensions remains a challenge. It has been proposed that they do so by becoming stable to diffeomorphisms, yet existing empirical measurements support it is often not the case. We revisit this question defining maximum-entropy distribution on allows study typical diffeomorphisms of given norm. confirm stability toward does strongly correlate performance benchmark sets images. By contrast, we find relative generic transformations R f correlates remarkably with test error ? t . order unity at initialization but decreases several decades during training for state-of-the-art architectures. For CIFAR10 and 15 known architectures ? t ? 0.2 R f , suggesting obtaining small important achieve good performance. how depends size set compare simple model invariant learning.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Stability of Anosov Diffeomorphisms

Definition 1. Let < , > be a C∞ Riemannian metric on M and | · | its induced norm on TxM for each x ∈ M . We say that f ∈ D is Anosov if 1. the tangent bundle of M splits in a Whitney direct sum of continuous subbundles TM = E ⊕ E, where E and E are Df -invariant, 2. there exists constants c, c′ > 0 and 0 < λ < 1 such that |Dfn xv| < c λn|v| |Df−n x w| < c′ λn|w| for all x ∈ M , v ∈ E x, and w ...

متن کامل

Deep Convolutional Neural Nets

The activation can be rewritten as follows a = vx+ b where v = [w1, . . . , wD] and b = wD+1 , and is an inner product between a weight vector v and the input x plus a bias b. For different inputs x of the same magnitude1, the activation is maximum when x is parallel to v, and the latter can be viewed as a pattern or template to which x is compared. The bias b then raises or lowers the activati...

متن کامل

Deep Fishing: Gradient Features from Deep Nets

Convolutional Networks (ConvNets) have recently improved image recognition performance thanks to end-to-end learning of deep feed-forward models from raw pixels. Deep learning is a marked departure from the previous state of the art, the Fisher Vector (FV), which relied on gradient-based encoding of local hand-crafted features. In this paper, we discuss a novel connection between these two appr...

متن کامل

Dynamics of surface diffeomorphisms relative to homoclinic and heteroclinic orbits

The Nielsen-Thurston theory of surface diffeomorphisms shows that useful dynamical information can be obtained from a finite collection of periodic orbits. In this paper, we extend these results to homoclinic and heteroclinic orbits of saddle points. These orbits are most readily computed and studied as intersections of unstable and stable manifolds comprising homoclinic or heteroclinic tangles...

متن کامل

Bigeometric Organization of Deep Nets

In this paper, we build an organization of high-dimensional datasets that cannot be cleanly embedded into a low-dimensional representation due to missing entries and a subset of the features being irrelevant to modeling functions of interest. Our algorithm begins by defining coarse neighborhoods of the points and defining an expected empirical function value on these neighborhoods. We then gene...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Journal of Statistical Mechanics: Theory and Experiment

سال: 2022

ISSN: ['1742-5468']

DOI: https://doi.org/10.1088/1742-5468/ac98ac